# Multimodal CLIP Architecture
Vit Base Patch16 Clip 224.laion400m E32
MIT
Vision Transformer model trained on the LAION-400M dataset, compatible with both open_clip and timm frameworks
Image Classification
V
timm
5,751
0
Vit Base Patch32 Clip 224.laion400m E31
MIT
Vision Transformer model trained on the LAION-400M dataset, compatible with both OpenCLIP and timm frameworks
Image Classification
V
timm
10.90k
0
Biomedclip Vit Bert Hf
MIT
A BiomedCLIP model implemented based on PyTorch and Huggingface frameworks, reproducing the original microsoft/BiomedCLIP-PubMedBERT_256-vit_base_patch16_224 model
Multimodal Fusion
Transformers English

B
chuhac
4,437
1
Featured Recommended AI Models